AIbase
Home
AI Tools
AI Models
MCP
AI NEWS
EN
Model Selection
Tags
GPTQ quantization

# GPTQ quantization

Qwen3 Embedding 4B W4A16 G128
Apache-2.0
This is the Qwen3-Embedding-4B model after GPTQ quantization, with significantly reduced video memory usage and minimal performance loss.
Text Embedding
Q
boboliu
141
1
Qwen3 Embedding 8B W4A16 G128
Apache-2.0
GPTQ quantized version of Qwen3-Embedding-8B, significantly reducing VRAM requirements while maintaining high performance
Text Embedding
Q
boboliu
322
1
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
English简体中文繁體中文にほんご
© 2025AIbase